Multimodal Adaptive Interfaces
نویسنده
چکیده
Our group is interested in creating human machine interfaces which use natural modalities such as vision and speech to sense and interpret a user's actions [6]. In this paper we describe recent work on multimodal adaptive interfaces which combine automatic speech recognition, computer vision for gesture tracking, and machine learning techniques. Speech is the primary mode of communication between people and should also be used in computer human communication. Gesture usually accompanies speech and provides information which is at times complementary and at times redundant to the information in the speech stream. Depending on the task at hand and the user's preferences, she will use a combination of speech and gesture in di erent ways to communicate her intent. In this paper we present preliminary results of an interface which lets the user communicate using a combination of speech and diectic (pointing) gestures. Although other e orts have been made to build multimodal interfaces, we present a system which centers around on-line learning to actively acquire communicative primitives from interactions with the user. In this paper we begin in Section 2 by examining some of the problems of designing interfaces which use natural modalities and motivate our approach which centers around enabling the interface to learn from the user. In Section 3 we give an overview of our approach to addressing the issues raised in Section 2. Section 4 describes the multimodal sensory environment we have built for developing our interfaces. This environment includes a vision based hand tracking system and a phonetic speech recognizer. In Section 5 we introduce Toco the Toucan, an animated synthetic character which provides embodiment for the interface. Section 6 details the learning algorithm which allows Toco to learn the acoustic models and meanings of words as the user points to virtual objects and talk about them. In Section 7 we summarize our work this far and nally in Section 8 we provide a road map for where this work is going.
منابع مشابه
Adaptive Mobile Multimodal User Interfaces for Dynamic Accessibility
In this paper we discuss how to exploit solutions able to support mobile multimodal adaptive user interfaces for dynamic accessibility. We focus on an approach based on the use of declarative user interface languages and oriented to Web applications accessed through emerging ubiquitous environments.
متن کاملHuman-centric framework for perceptually adaptive interfaces
Multimodal interfaces have long held the promise of enhanced and effective human machine interaction. The ultimate goal of multimodal interfaces is to facilitate human activity allowing seamless exchange of information. This goal requires a coordinated development effort that incorporates a thorough understanding of human perceptual system in the design of interfaces. In this manner, multimodal...
متن کاملAdaptive multimodal web user interfaces for smart work environments
This paper presents a solution for supporting adaptive user interfaces in work environments that require operators to move about in dynamic contexts for manipulating various physical objects. The solution architecture is built upon the use of logical languages for interactive applications integrated with context aware and adaptive features. The proposed architecture is able to adapt to specific...
متن کاملTangible Objects for the Acquisition of Multimodal Interaction Patterns
Multimodal user interfaces offer more intuitive interaction for end-users, however, usually only through predefined input schemes. This paper describes a user experiment for multimodal interaction pattern identification, using head gesture and speech inputs for a 3D graph manipulation. We show that a direct mapping between head gestures and the 3D object predominates, however even for such a si...
متن کاملSemantic Modelling in Support of Adaptive Multimodal Interface Design
The design of multimodal interfaces requires intelligent data interpretation in order to guarantee seamless adaptation to the user’s needs and context. HMI (humanmachine interaction) design accommodates varying forms of interaction patterns, depending on what is most appropriate for a particular user at a particular time. These design patterns are a powerful means of documenting reusable design...
متن کاملA Cross-Layer Design for Adaptive Multimodal Interfaces in Pervasive Computing
Multimodal interfaces have attracted more and more attention. Most researches focus on each communication mode independently and then fuse the information at the application level. Recently, several frameworks and models have been proposed to support the design and development of multimodal interfaces. However, it is still a challenging issue of supporting adaptations in the multimodal interfac...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1997